Tootfinder

Opt-in global Mastodon full text search. Join the index!

@arXiv_csCL_bot@mastoxiv.page
2024-03-12 06:48:50

SCORE: Self-supervised Correspondence Fine-tuning for Improved Content Representations
Amit Meghanani, Thomas Hain
arxiv.org/abs/2403.06260

@arXiv_csSE_bot@mastoxiv.page
2024-04-12 07:34:55

Structure-aware Fine-tuning for Code Pre-trained Models
Jiayi Wu, Renyu Zhu, Nuo Chen, Qiushi Sun, Xiang Li, Ming Gao
arxiv.org/abs/2404.07471

@arXiv_mathOC_bot@mastoxiv.page
2024-03-12 06:58:23

Fine-tuning of diffusion models via stochastic control: entropy regularization and beyond
Wenpin Tang
arxiv.org/abs/2403.06279

@arXiv_csLG_bot@mastoxiv.page
2024-02-12 08:33:50

This arxiv.org/abs/2312.14378 has been replaced.
initial toot: mastoxiv.page/@arXiv_csLG_…

@Mediagazer@mstdn.social
2024-03-10 13:40:23

William Whitworth, who wrote revealing profiles in The New Yorker before working as the EIC at The Atlantic from 1980 to 1999, died on March 8 at age 87 (Sam Roberts/New York Times)
nytimes.com/2024/03/09/…

Killings by police brought reforms. Fear of crime is unraveling them.
There is a groundswell of legislative and voter #pushback AGAINST #reforms initiated over the past four years after the #police

@arXiv_csSE_bot@mastoxiv.page
2024-02-12 06:53:02

Delving into Parameter-Efficient Fine-Tuning in Code Change Learning: An Empirical Study
Shuo Liu, Jacky Keung, Zhen Yang, Fang Liu, Qilin Zhou, Yihan Liao
arxiv.org/abs/2402.06247

@arXiv_csDC_bot@mastoxiv.page
2024-03-12 07:17:50

Adding NVMe SSDs to Enable and Accelerate 100B Model Fine-tuning on a Single GPU
Changyue Liao, Mo Sun, Zihan Yang, Kaiqi Chen, Binhang Yuan, Fei Wu, Zeke Wang
arxiv.org/abs/2403.06504

@arXiv_csCV_bot@mastoxiv.page
2024-02-12 08:31:17

This arxiv.org/abs/2401.05126 has been replaced.
initial toot: mastoxiv.page/@arXiv_csCV_…

@arXiv_csCY_bot@mastoxiv.page
2024-03-12 07:16:15

Improving Low-Resource Knowledge Tracing Tasks by Supervised Pre-training and Importance Mechanism Fine-tuning
Hengyuan Zhang, Zitao Liu, Shuyan Huang, Chenming Shang, Bojun Zhan, Yong Jiang
arxiv.org/abs/2403.06725

@arXiv_csNE_bot@mastoxiv.page
2024-02-12 07:25:14

Fine-Tuning Surrogate Gradient Learning for Optimal Hardware Performance in Spiking Neural Networks
Ilkin Aliyev, Tosiron Adegbija
arxiv.org/abs/2402.06211

@SmartmanApps@dotnet.social
2024-04-10 10:58:56

#ExplainVintageTechnology
#HashTagGames
It's a TV... but with no colour, the screen was curved, and you had to fiddle endlessly with the rabbit's ears until you finally got a good picture. The you had to twist a dial called "fine tuning" to try and get…

@arXiv_csCL_bot@mastoxiv.page
2024-03-12 06:48:55

IndicLLMSuite: A Blueprint for Creating Pre-training and Fine-Tuning Datasets for Indian Languages
Mohammed Safi Ur Rahman Khan, Priyam Mehta, Ananth Sankar, Umashankar Kumaravelan, Sumanth Doddapaneni, Suriyaprasaad G, Varun Balan G, Sparsh Jain, Anoop Kunchukuttan, Pratyush Kumar, Raj Dabre, Mitesh M. Khapra
arxiv.org/…

@arXiv_csLG_bot@mastoxiv.page
2024-04-10 06:51:47

Automated Federated Pipeline for Parameter-Efficient Fine-Tuning of Large Language Models
Zihan Fang, Zheng Lin, Zhe Chen, Xianhao Chen, Yue Gao, Yuguang Fang
arxiv.org/abs/2404.06448

@arXiv_csCR_bot@mastoxiv.page
2024-04-09 06:48:00

Increased LLM Vulnerabilities from Fine-tuning and Quantization
Divyanshu Kumar, Anurakt Kumar, Sahil Agarwal, Prashanth Harshangi
arxiv.org/abs/2404.04392

@arXiv_csGT_bot@mastoxiv.page
2024-05-10 06:50:09

Truthful Aggregation of LLMs with an Application to Online Advertising
Ermis Soumalias, Michael J. Curry, Sven Seuken
arxiv.org/abs/2405.05905

@arXiv_csHC_bot@mastoxiv.page
2024-02-12 07:15:55

Randomness Is All You Need: Semantic Traversal of Problem-Solution Spaces with Large Language Models
Thomas Sandholm, Sayandev Mukherjee, Bernardo A. Huberman
arxiv.org/abs/2402.06053

@arXiv_csSE_bot@mastoxiv.page
2024-03-12 08:42:05

This arxiv.org/abs/2312.15698 has been replaced.
link: scholar.google.com/scholar?q=a

@arXiv_csCL_bot@mastoxiv.page
2024-03-11 08:30:05

This arxiv.org/abs/2309.12307 has been replaced.
initial toot: mastoxiv.page/@arXiv_csCL_…

@arXiv_physicsoptics_bot@mastoxiv.page
2024-04-12 07:07:06

Si Superstrate Lenses on Patch-Antenna-Coupled TeraFETs: NEP Optimization and Frequency Fine-Tuning
Anastasiya Krysl, Dmytro B. But, K\k{e}stutis Ikamas, Jakob Holstein, Anna Shevchik-Shekera, Hartmut G. Roskos, Alvydas Lisauskas
arxiv.org/abs/2404.07715

@arXiv_csCV_bot@mastoxiv.page
2024-04-12 08:30:33

This arxiv.org/abs/2404.05426 has been replaced.
initial toot: mastoxiv.page/@arXiv_csCV_…

@arXiv_csCL_bot@mastoxiv.page
2024-03-12 06:49:05

'One size doesn't fit all': Learning how many Examples to use for In-Context Learning for Improved Text Classification
Manish Chandra, Debasis Ganguly, Yiwen Li, Iadh Ounis
arxiv.org/abs/2403.06402

@TedUnderwood@sigmoid.social
2024-05-06 13:37:23

Jack Clark’s Import AI newsletter is always fun; this one especially because it explores the shared interest Meta and the CCP have in preventing fine-tuning of their models.
In a “Palantír”-level irony, one of the projects to prevent all further learning is called SOPHON.

@arXiv_physicsplasmph_bot@mastoxiv.page
2024-03-12 07:22:41

Data-driven sparse modeling of oscillations in plasma space propulsion
B. Bay\'on-Buj\'an, M. Merino
arxiv.org/abs/2403.06809

@arXiv_condmatmtrlsci_bot@mastoxiv.page
2024-04-11 07:25:18

SAM-I-Am: Semantic Boosting for Zero-shot Atomic-Scale Electron Micrograph Segmentation
Waqwoya Abebe, Jan Strube, Luanzheng Guo, Nathan R. Tallent, Oceane Bel, Steven Spurgeon, Christina Doty, Ali Jannesari
arxiv.org/abs/2404.06638

@arXiv_grqc_bot@mastoxiv.page
2024-04-08 07:01:59

Geometry from geodesics: fine-tuning Ehlers, Pirani, and Schild
James T. Wheeler
arxiv.org/abs/2404.03815 arxiv.org/p…

@Techmeme@techhub.social
2024-04-04 15:51:04

OpenAI expands its Custom Model training program with "assisted fine-tuning", letting organizations set up data training pipelines, evaluation systems, and more (Kyle Wiggers/TechCrunch)
techcrunch.com/2024/04/04/open

@arXiv_csCL_bot@mastoxiv.page
2024-02-12 07:33:33

Aya Dataset: An Open-Access Collection for Multilingual Instruction Tuning
Shivalika Singh, Freddie Vargus, Daniel Dsouza, B\"orje F. Karlsson, Abinaya Mahendiran, Wei-Yin Ko, Herumb Shandilya, Jay Patel, Deividas Mataciunas, Laura OMahony, Mike Zhang, Ramith Hettiarachchi, Joseph Wilson, Marina Machado, Luisa Souza Moura, Dominik Krzemi\'nski, Hakimeh Fadaei, Irem Erg\"un, Ifeoma Okoh, Aisha Alaagib, Oshan Mudannayake, Zaid Alyafeai, Vu Minh Chien, Sebastian Ruder, Surya…

@jgkoomey@mastodon.energy
2024-03-24 20:30:50

Speaking of vaccines, this could be very important. mag.uchicago.edu/science-medic

@mapto@qoto.org
2024-05-08 04:36:44

@… @… actually, I do believe you can remix models that are openly shared (at least through fine-tuning).
But anything closed that uses SO (e.g. Google BigQuery), might probably be an infringement.

@arXiv_csLG_bot@mastoxiv.page
2024-02-12 08:34:08

This arxiv.org/abs/2402.04004 has been replaced.
link: scholar.google.com/scholar?q=a

@arXiv_hepph_bot@mastoxiv.page
2024-05-10 08:41:01

This arxiv.org/abs/2310.06611 has been replaced.
initial toot: mastoxiv.page/@arXiv_hepp…

@arXiv_csCV_bot@mastoxiv.page
2024-05-10 08:29:21

This arxiv.org/abs/2312.03045 has been replaced.
initial toot: mastoxiv.page/@arXiv_csCV_…

@arXiv_eessIV_bot@mastoxiv.page
2024-03-07 07:28:27

Low-Dose CT Image Reconstruction by Fine-Tuning a UNet Pretrained for Gaussian Denoising for the Downstream Task of Image Enhancement
Tim Selig, Thomas M\"arz, Martin Storath, Andreas Weinmann
arxiv.org/abs/2403.03551

@arXiv_csNE_bot@mastoxiv.page
2024-04-12 06:51:16

AD-NEv : The multi-architecture neuroevolution-based multivariate anomaly detection framework
Marcin Pietro\'n, Dominik \.Zurek, Kamil Faber, Roberto Corizzo
arxiv.org/abs/2404.07968

@arXiv_eessAS_bot@mastoxiv.page
2024-04-11 08:34:56

This arxiv.org/abs/2307.07218 has been replaced.
initial toot: mastoxiv.page/@arXiv_ees…

@arXiv_csSD_bot@mastoxiv.page
2024-04-09 08:47:55

This arxiv.org/abs/2402.10100 has been replaced.
initial toot: mastoxiv.page/@arXiv_csSD_…

@arXiv_csCV_bot@mastoxiv.page
2024-02-12 07:04:47

Iris-SAM: Iris Segmentation Using a Foundational Model
Parisa Farmanifard, Arun Ross
arxiv.org/abs/2402.06497 arxiv.o…

@arXiv_astrophCO_bot@mastoxiv.page
2024-03-04 08:39:30

This arxiv.org/abs/2401.04190 has been replaced.
initial toot: mastoxiv.page/@arXiv_…

@arXiv_csSE_bot@mastoxiv.page
2024-03-12 08:41:32

This arxiv.org/abs/2305.00418 has been replaced.
initial toot: mastoxiv.page/@arXiv_csSE_…

@arXiv_csIR_bot@mastoxiv.page
2024-03-28 08:29:20

This arxiv.org/abs/2403.16915 has been replaced.
initial toot: mastoxiv.page/@arXiv_csIR_…

@arXiv_csCR_bot@mastoxiv.page
2024-05-10 06:48:07

PLLM-CS: Pre-trained Large Language Model (LLM) for Cyber Threat Detection in Satellite Networks
Mohammed Hassanin, Marwa Keshk, Sara Salim, Majid Alsubaie, Dharmendra Sharma
arxiv.org/abs/2405.05469

@williamgunn@mastodon.social
2024-04-02 15:52:12

Wiley licenses content for training an #LLM. The company was not named, but I would suspect it's the one which has been signing a lot of licensing deals lately. Access to STM content could be a big differentiator, though I wouldn't expect it to be exclusive. Also, $23M sounds small.

A screenshot reading, "As we discussed on our investor update, our content is foundational for training and fine-tuning these models. I'm pleased to report that after the quarter closed, we executed a $23 million content rights project with a large tech company. The onetime transaction to be recorded in Q4 includes access to previously published academic and professional book content for specific use in training LLM models. We are working to uncover similar content opportunities with other AI p…
@arXiv_csCL_bot@mastoxiv.page
2024-04-11 08:30:25

This arxiv.org/abs/2310.05910 has been replaced.
link: scholar.google.com/scholar?q=a

@arXiv_eessSP_bot@mastoxiv.page
2024-05-08 08:38:16

This arxiv.org/abs/2405.02422 has been replaced.
initial toot: mastoxiv.page/@arXiv_ees…

@arXiv_hepph_bot@mastoxiv.page
2024-04-10 07:02:19

Higgs Alignment from Multicritical-Point Principle in Two Higgs Doublet Models
Hikaru Kawai, Kiyoharu Kawana, Kin-ya Oda, Kei Yagyu
arxiv.org/abs/2404.06096

@arXiv_csCL_bot@mastoxiv.page
2024-03-11 08:30:21

This arxiv.org/abs/2312.09979 has been replaced.
initial toot: mastoxiv.page/@arXiv_csCL_…

@arXiv_csSE_bot@mastoxiv.page
2024-05-10 08:32:38

This arxiv.org/abs/2306.07285 has been replaced.
initial toot: mastoxiv.page/@arXiv_csSE_…

@arXiv_csCL_bot@mastoxiv.page
2024-03-11 07:18:49

Deep Prompt Multi-task Network for Abuse Language Detection
Jian Zhu, Yuping Ruan, Jingfei Chang, Cheng Luo
arxiv.org/abs/2403.05268

@arXiv_astrophCO_bot@mastoxiv.page
2024-03-04 08:39:30

This arxiv.org/abs/2401.04190 has been replaced.
initial toot: mastoxiv.page/@arXiv_…

@arXiv_eessIV_bot@mastoxiv.page
2024-05-09 08:34:53

This arxiv.org/abs/2404.15786 has been replaced.
link: scholar.google.com/scholar?q=a

@arXiv_csCV_bot@mastoxiv.page
2024-04-10 07:35:33

Magic-Boost: Boost 3D Generation with Mutli-View Conditioned Diffusion
Fan Yang, Jianfeng Zhang, Yichun Shi, Bowen Chen, Chenxu Zhang, Huichao Zhang, Xiaofeng Yang, Jiashi Feng, Guosheng Lin
arxiv.org/abs/2404.06429

@arXiv_csCL_bot@mastoxiv.page
2024-03-11 07:18:49

Deep Prompt Multi-task Network for Abuse Language Detection
Jian Zhu, Yuping Ruan, Jingfei Chang, Cheng Luo
arxiv.org/abs/2403.05268

@arXiv_csIR_bot@mastoxiv.page
2024-03-26 06:50:36

Coarse-Tuning for Ad-hoc Document Retrieval Using Pre-trained Language Models
Atsushi Keyaki, Ribeka Keyaki
arxiv.org/abs/2403.16915

@arXiv_csCL_bot@mastoxiv.page
2024-03-11 07:18:46

Cross-lingual Transfer or Machine Translation? On Data Augmentation for Monolingual Semantic Textual Similarity
Sho Hoshino, Akihiko Kato, Soichiro Murakami, Peinan Zhang
arxiv.org/abs/2403.05257

@arXiv_csCR_bot@mastoxiv.page
2024-02-26 06:47:59

Mitigating Fine-tuning Jailbreak Attack with Backdoor Enhanced Alignment
Jiongxiao Wang, Jiazhao Li, Yiquan Li, Xiangyu Qi, Muhao Chen, Junjie Hu, Yixuan Li, Bo Li, Chaowei Xiao
arxiv.org/abs/2402.14968

@arXiv_csCL_bot@mastoxiv.page
2024-03-11 07:18:46

Cross-lingual Transfer or Machine Translation? On Data Augmentation for Monolingual Semantic Textual Similarity
Sho Hoshino, Akihiko Kato, Soichiro Murakami, Peinan Zhang
arxiv.org/abs/2403.05257

@arXiv_csCV_bot@mastoxiv.page
2024-04-10 07:35:46

Can Feedback Enhance Semantic Grounding in Large Vision-Language Models?
Yuan-Hong Liao, Rafid Mahmood, Sanja Fidler, David Acuna
arxiv.org/abs/2404.06510

@arXiv_csNE_bot@mastoxiv.page
2024-04-09 08:46:16

This arxiv.org/abs/2304.04067 has been replaced.
initial toot: mastoxiv.page/@arXiv_csNE_…

@arXiv_csCL_bot@mastoxiv.page
2024-02-12 08:30:24

This arxiv.org/abs/2305.18582 has been replaced.
initial toot: mastoxiv.page/@arXiv_csCL_…

@arXiv_csCL_bot@mastoxiv.page
2024-02-12 08:30:24

This arxiv.org/abs/2305.18582 has been replaced.
initial toot: mastoxiv.page/@arXiv_csCL_…

@arXiv_csSE_bot@mastoxiv.page
2024-04-03 06:52:52

FT2Ra: A Fine-Tuning-Inspired Approach to Retrieval-Augmented Code Completion
Qi Guo, Xiaohong Li, Xiaofei Xie, Shangqing Liu, Ze Tang, Ruitao Feng, Junjie Wang, Jidong Ge, Lei Bu
arxiv.org/abs/2404.01554

@arXiv_csLG_bot@mastoxiv.page
2024-04-30 09:09:07

This arxiv.org/abs/2404.11536 has been replaced.
link: scholar.google.com/scholar?q=a

@arXiv_csCR_bot@mastoxiv.page
2024-02-29 08:31:40

This arxiv.org/abs/2402.14968 has been replaced.
initial toot: mastoxiv.page/@arXiv_csCR_…

@arXiv_csIR_bot@mastoxiv.page
2024-03-27 08:25:29

This arxiv.org/abs/2403.16915 has been replaced.
initial toot: mastoxiv.page/@arXiv_csIR_…

@arXiv_csCL_bot@mastoxiv.page
2024-03-11 07:19:05

Bias-Augmented Consistency Training Reduces Biased Reasoning in Chain-of-Thought
James Chua, Edward Rees, Hunar Batra, Samuel R. Bowman, Julian Michael, Ethan Perez, Miles Turpin
arxiv.org/abs/2403.05518

@arXiv_csCL_bot@mastoxiv.page
2024-03-11 07:19:05

Bias-Augmented Consistency Training Reduces Biased Reasoning in Chain-of-Thought
James Chua, Edward Rees, Hunar Batra, Samuel R. Bowman, Julian Michael, Ethan Perez, Miles Turpin
arxiv.org/abs/2403.05518

@arXiv_csCL_bot@mastoxiv.page
2024-02-12 08:30:22

This arxiv.org/abs/2305.13179 has been replaced.
link: scholar.google.com/scholar?q=a

@arXiv_csNE_bot@mastoxiv.page
2024-05-09 06:52:55

Learning-to-learn enables rapid learning with phase-change memory-based in-memory computing
Thomas Ortner, Horst Petschenig, Athanasios Vasilopoulos, Roland Renner, \v{S}pela Brglez, Thomas Limbacher, Enrique Pi\~nero, Alejandro Linares Barranco, Angeliki Pantazi, Robert Legenstein
arxiv.org/abs/2405.05141

@arXiv_csSE_bot@mastoxiv.page
2024-05-06 07:25:00

Empirical Studies of Parameter Efficient Methods for Large Language Models of Code and Knowledge Transfer to R
Amirreza Esmaeili, Iman Saberi, Fatemeh H. Fard
arxiv.org/abs/2405.01553

@arXiv_csCL_bot@mastoxiv.page
2024-02-12 08:30:22

This arxiv.org/abs/2305.13179 has been replaced.
link: scholar.google.com/scholar?q=a

@arXiv_csCL_bot@mastoxiv.page
2024-04-08 06:48:27

Unlocking Parameter-Efficient Fine-Tuning for Low-Resource Language Translation
Tong Su, Xin Peng, Sarubi Thillainathan, David Guzm\'an, Surangika Ranathunga, En-Shiun Annie Lee
arxiv.org/abs/2404.04212

@arXiv_csIR_bot@mastoxiv.page
2024-02-28 06:50:23

A Fine-tuning Enhanced RAG System with Quantized Influence Measure as AI Judge
Keshav Rangan, Yiqiao Yin
arxiv.org/abs/2402.17081

@arXiv_csCR_bot@mastoxiv.page
2024-04-01 08:29:56

This arxiv.org/abs/2402.12168 has been replaced.
initial toot: mastoxiv.page/@arXiv_csCR_…

@arXiv_csLG_bot@mastoxiv.page
2024-04-24 08:49:22

This arxiv.org/abs/2404.14367 has been replaced.
initial toot: mastoxiv.page/@arXiv_csLG_…

@arXiv_csCL_bot@mastoxiv.page
2024-05-09 06:48:50

Fine-tuning Pre-trained Named Entity Recognition Models For Indian Languages
Sankalp Bahad, Pruthwik Mishra, Karunesh Arora, Rakesh Chandra Balabantaray, Dipti Misra Sharma, Parameswari Krishnamurthy
arxiv.org/abs/2405.04829

@arXiv_csCL_bot@mastoxiv.page
2024-05-09 06:48:50

Fine-tuning Pre-trained Named Entity Recognition Models For Indian Languages
Sankalp Bahad, Pruthwik Mishra, Karunesh Arora, Rakesh Chandra Balabantaray, Dipti Misra Sharma, Parameswari Krishnamurthy
arxiv.org/abs/2405.04829

@arXiv_csSE_bot@mastoxiv.page
2024-04-09 06:53:22

Constraining Large Language Model for Generating Computer-Parsable Content
Jiaye Wang
arxiv.org/abs/2404.05499 arxiv.…

@arXiv_csCR_bot@mastoxiv.page
2024-02-27 08:19:06

This arxiv.org/abs/2310.09266 has been replaced.
initial toot: mastoxiv.page/@arXiv_csCR_…

@arXiv_csCV_bot@mastoxiv.page
2024-02-29 08:34:37

This arxiv.org/abs/2402.17412 has been replaced.
initial toot: mastoxiv.page/@arXiv_csCV_…

@arXiv_csCL_bot@mastoxiv.page
2024-03-08 08:29:41

This arxiv.org/abs/2403.01432 has been replaced.
link: scholar.google.com/scholar?q=a

@arXiv_csSE_bot@mastoxiv.page
2024-05-03 08:47:43

This arxiv.org/abs/2402.00905 has been replaced.
link: scholar.google.com/scholar?q=a

@arXiv_csCL_bot@mastoxiv.page
2024-05-03 07:15:29

LoRA Land: 310 Fine-tuned LLMs that Rival GPT-4, A Technical Report
Justin Zhao, Timothy Wang, Wael Abid, Geoffrey Angus, Arnav Garg, Jeffery Kinnison, Alex Sherstinsky, Piero Molino, Travis Addair, Devvret Rishi
arxiv.org/abs/2405.00732

@arXiv_csSE_bot@mastoxiv.page
2024-04-22 06:52:51

Multi-Objective Fine-Tuning for Enhanced Program Repair with LLMs
Boyang Yang, Haoye Tian, Jiadong Ren, Hongyu Zhang, Jacques Klein, Tegawend\'e F. Bissyand\'e, Claire Le Goues, Shunfu Jin
arxiv.org/abs/2404.12636

@arXiv_csSE_bot@mastoxiv.page
2024-05-02 08:30:20

This arxiv.org/abs/2402.00905 has been replaced.
link: scholar.google.com/scholar?q=a

@arXiv_csCL_bot@mastoxiv.page
2024-04-05 08:31:09

This arxiv.org/abs/2403.20145 has been replaced.
initial toot: mastoxiv.page/@arXiv_csCL_…

@arXiv_csCL_bot@mastoxiv.page
2024-02-23 06:57:17

Fine-Tuning Enhances Existing Mechanisms: A Case Study on Entity Tracking
Nikhil Prakash, Tamar Rott Shaham, Tal Haklay, Yonatan Belinkov, David Bau
arxiv.org/abs/2402.14811

@arXiv_csCL_bot@mastoxiv.page
2024-05-06 08:26:43

This arxiv.org/abs/2403.09891 has been replaced.
initial toot: mastoxiv.page/@arXiv_csCL_…

@arXiv_csCL_bot@mastoxiv.page
2024-03-01 06:53:46

OpenMedLM: Prompt engineering can out-perform fine-tuning in medical question-answering with open-source large language models
Jenish Maharjan, Anurag Garikipati, Navan Preet Singh, Leo Cyrus, Mayank Sharma, Madalina Ciobanu, Gina Barnes, Rahul Thapa, Qingqing Mao, Ritankar Das
arxiv.org/abs/2402.19371

@arXiv_csCL_bot@mastoxiv.page
2024-03-04 07:27:14

Cross-Lingual Learning vs. Low-Resource Fine-Tuning: A Case Study with Fact-Checking in Turkish
Recep Firat Cekinel, Pinar Karagoz, Cagri Coltekin
arxiv.org/abs/2403.00411

@arXiv_csCL_bot@mastoxiv.page
2024-02-29 06:50:44

Learning or Self-aligning? Rethinking Instruction Fine-tuning
Mengjie Ren, Boxi Cao, Hongyu Lin, Liu Cao, Xianpei Han, Ke Zeng, Guanglu Wan, Xunliang Cai, Le Sun
arxiv.org/abs/2402.18243

@arXiv_csCL_bot@mastoxiv.page
2024-04-04 08:33:51

This arxiv.org/abs/2404.00213 has been replaced.
link: scholar.google.com/scholar?q=a

@arXiv_csCL_bot@mastoxiv.page
2024-04-08 06:48:04

Simple Techniques for Enhancing Sentence Embeddings in Generative Language Models
Bowen Zhang, Kehua Chang, Chunping Li
arxiv.org/abs/2404.03921

@arXiv_csCL_bot@mastoxiv.page
2024-03-07 08:24:50

This arxiv.org/abs/2309.13734 has been replaced.
link: scholar.google.com/scholar?q=a

@arXiv_csCL_bot@mastoxiv.page
2024-02-29 06:50:58

Is Crowdsourcing Breaking Your Bank? Cost-Effective Fine-Tuning of Pre-trained Language Models with Proximal Policy Optimization
Shuo Yang, Gjergji Kasneci
arxiv.org/abs/2402.18284

@arXiv_csCL_bot@mastoxiv.page
2024-03-29 08:32:05

This arxiv.org/abs/2403.18025 has been replaced.
initial toot: mastoxiv.page/@arXiv_csCL_…

@arXiv_csCL_bot@mastoxiv.page
2024-02-23 06:55:57

INSTRAUG: Automatic Instruction Augmentation for Multimodal Instruction Fine-tuning
Wei Han, Hui Chen, Soujanya Poria
arxiv.org/abs/2402.14492